Cooperative Batch Scheduling for HPC Systems

نویسندگان

  • Xu Yang
  • Zhiling Lan
چکیده

The batch scheduler is an important system software serving as the interface between users and HPC systems. Users submit their jobs via batch scheduling portal and the batch scheduler makes scheduling decision for each job based on its request for computing sources, i.e. core-hours. However, jobs submitted to HPC systems are usually parallel applications and their lifecycle consists of multiple running phases, such as computation, communication and I/O. Thus, the running of a job could involve different kinds of system resources, such as power, network bandwidth, I/O bandwidth, storage, etc. Today’s batch schedulers rarely take these resource requirements into consideration for making scheduling decisions, which has been identified as one of the major culprits for system wide

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Comparison of Job Management Systems in Supporting HPC ClusterTools

This paper compares three most common job management systems and their workings with Sun HPC ClusterTools 3.1. Various aspects such as installation, customization, scheduling and resource control issues are discussed. The three chosen systems are: Load Sharing Facility (LSF), Portable Batch System (PBS) and COmputing in DIstributed Networked Environment (CODINE)/ Global Resource Director (GRD)....

متن کامل

myHadoop - Hadoop-on-Demand on Traditional HPC Resources

Traditional High Performance Computing (HPC) resources, such as those available on the TeraGrid, support batch job submissions using Distributed Resource Management Systems (DRMS) like TORQUE or the Sun Grid Engine (SGE). For large-scale data intensive computing, programming paradigms such as MapReduce are becoming popular. A growing number of codes in scientific domains such as Bioinformatics ...

متن کامل

A note on new trends in data-aware scheduling and resource provisioning in modern HPC systems

The Big Data era [1,2] poses new challenges as well as significant opportunities for High-Performance Computing (HPC) systems such as how to efficiently turn massively large data into valuable information and meaningful knowledge? It is clear that computationally optimized new data-driven HPC techniques are required for processing Big Data in rapidly-increasing number of applications, such as L...

متن کامل

Ac 2007-2445: High-performance Computing Student Projects

Commodity High Performance Computing (HPC) platforms such as Beowulf Clusters provide excellent opportunities to engage students with challenging projects. Courses such as parallel programming, distributed systems, operating systems, and networking can benefit from the lowcost HPC platform. In this paper we report the results on series of student projects in an advanced operating systems course...

متن کامل

Dynamic Fractional Resource Scheduling vs. Batch Scheduling

We propose a novel job scheduling approach for homogeneous cluster computing platforms. Its key feature is the use of virtual machine technology to share fractional node resources in a precise and controlled manner. Other VM-based scheduling approaches have focused primarily on technical issues or on extensions to existing batch scheduling systems, while we take a more aggressive approach and s...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016